Computing Sparse Representations in O(N log N) Time

نویسنده

  • Tsung-Han Lin
چکیده

Machine learning concerns forming representations of input observations to facilitate tasks such as classification. A recent insight in deep learning [1] is to use a deep architecture that stacks multiple levels of nonlinear operations in an inference hierarchy to extract different layers of abstractions. Deep learning is a promising direction and has attained state-of-the-art performance in some application areas such as computer vision and speech recognition. In this work, we aim to improve the computational efficiency of deep learning. A key step in deep learning is to represent input signals as layers of sparse representations. In image processing, this means to progressively describe objects using features of larger spatial scales. For example, in the bottom layer objects can be represented by edges of different widths, lengths and orientations over small regions. In higher layers objects may be represented by shapes such as squares, triangles and so on over large regions. The representations are computed based on some designed or learned dictionary. At each layer an input signal is represented using just a few dictionary atoms. We consider the orthogonal matching pursuit (OMP) algorithm [5], which forms sparse representations by greedily selecting representing dictionary atoms. The computation cost of OMP is proportional to the dimensions of the dictionary. Suppose that the input signal is an M×1 vector x, and the dictionary has N M×1 atoms di, i = 1, ..., N . The bulk of the sparse representation computation amounts to computing N correlations between x and di for all i. Thus the total computation cost is O(MN). Note that N is usually governed by the characteristics of a given machine learning task. For example, if the task is to classify objects with a large number of categories, N tends to be large for an increased chance of representing x with just a few dictionary atoms. On the other hand, with an easier task such as differentiating between only a few different looking objects, a relatively small N may be sufficient to derive discernible representations. M , however, is driven by the input signal size and the size of intermediate sparse representations computed in the hierarchy. This means M can be large especially in higher layers of the learning framework. For example, in image processing, sparse representations for small local regions of an image are formed in the bottom layer. The representations are then aggregated and vectorized over a larger neighborhood as the input signal for the next layer, which can easily be very long. We show that this O(MN) cost can be reduced to O(N logN), a complexity independent of the signal or representation size M . This means that the computation cost is only dictated by the desired classification resolution.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Eecient Computation of Implicit Representations of Sparse Graphs (revised Version) Eecient C O M Putation of Implicit Representations of Sparse Graphs

The problem of nding an implicit representation for a graph such that vertex ad jacency can be tested quickly is fundamental to all graph algorithms In particular it is possible to represent sparse graphs on n vertices using O n space such that vertex adjacency is tested in O time We show here how to construct such a representation e ciently by providing simple and optimal algorithms both in a ...

متن کامل

Efficient Computation of Implicit Representations of Sparse Graphs

The problem of nding an implicit representation for a graph such that vertex ad-jacency can be tested quickly is fundamental to all graph algorithms. In particular, it is possible to represent sparse graphs on n vertices using O(n) space such that vertex adjacency is tested in O(1) time. We show here how to construct such a representation eeciently by providing simple and optimal algorithms, bo...

متن کامل

(Nearly) Sample-Optimal Sparse Fourier Transform

We consider the problem of computing a k-sparse approximation to the discrete Fourier transform of an n-dimensional signal. Our main result is a randomized algorithm that computes such an approximation using O(k log n(log log n)) signal samples in time O(k log n(log log n)), assuming that the entries of the signal are polynomially bounded. The sampling complexity improves over the recent bound ...

متن کامل

(2∆− 1)-Edge-Coloring is Much Easier than Maximal Matching in Distributed Setting

Graph coloring is a central problem in distributed computing. Both vertexand edge-coloring problems have been extensively studied in this context. In this paper we show that a (2∆ − 1)-edge-coloring can be computed in time smaller than log n for any > 0, specifically, in e √ log logn) rounds. This establishes a separation between the (2∆− 1)-edge-coloring and the Maximal Matching problems, as t...

متن کامل

A robust sub-linear time R-FFAST algorithm for computing a sparse DFT

The Fast Fourier Transform (FFT) is the most efficiently known way to compute the Discrete Fourier Transform (DFT) of an arbitrary n-length signal, and has a computational complexity of O(n log n). If the DFT ~ X of the signal ~x has only k non-zero coefficients (where k < n), can we do better? In [1], we addressed this question and presented a novel FFAST (Fast Fourier Aliasing-based Sparse Tr...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2013